Bayesian Reasoning with Trained Neural Networks
نویسندگان
چکیده
منابع مشابه
Recurrent Bayesian Reasoning in Probabilistic Neural Networks
Considering the probabilistic approach to neural networks in the framework of statistical pattern recognition we assume approximation of class-conditional probability distributions by finite mixtures of product components. The mixture components can be interpreted as probabilistic neurons in neurophysiological terms and, in this respect, the fixed probabilistic description becomes conflicting w...
متن کاملGraphical Reasoning with Bayesian Networks
Nowadays, Bayesian networks are seen by many researchers as standard tools for reasoning with uncertainty. Despite the fact that Bayesian networks are graphical representations, representing dependence and independence information, normally the emphasis of the visualisation of the reasoning process is on showing changes in the associated marginal probability distributions due to entering observ...
متن کاملInstantaneously Trained Neural Networks
This paper presents a review of instantaneously trained neural networks (ITNNs). These networks trade learning time for size and, in the basic model, a new hidden node is created for each training sample. Various versions of the cornerclassification family of ITNNs, which have found applications in artificial intelligence (AI), are described. Implementation issues are also considered.
متن کاملBayesian Case - Based Reasoning with Neural
| Given a problem, a case-based reasoning (CBR) system will search its case memory and use the stored cases to nd the solution, possibly modifying retrieved cases to adapt to the required input speciications. In this paper we introduce a neural network architecture for eecient case-based reasoning. We show how a rigorous Bayesian probability propagation algorithm can be implemented as a feedfor...
متن کاملBayesian Optimization with Robust Bayesian Neural Networks
Bayesian optimization is a prominent method for optimizing expensive-to-evaluate black-box functions that is widely applied to tuning the hyperparameters of machine learning algorithms. Despite its successes, the prototypical Bayesian optimization approach – using Gaussian process models – does not scale well to either many hyperparameters or many function evaluations. Attacking this lack of sc...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Entropy
سال: 2021
ISSN: 1099-4300
DOI: 10.3390/e23060693